我们在盒子值上相关的分布中重新审视经典的潘多拉盒(PB)问题。 ARXIV的最新工作:1911.01632获得了限制性类别的策略量持续近似算法,该策略以固定顺序访问框。在这项工作中,我们研究了近似最佳策略的复杂性,该策略可以根据迄今为止所看到的值适应下一步访问哪个框。我们的主要结果确定了PB的近似值等效性与研究良好的统一决策树(UDT)问题,从随机优化和Min-Sum Set封面的变体($ \ MATHCAL {MSSC} _F $)问题。对于支持$ M $的分布,UDT承认$ \ log M $近似值,而多项式时间的恒定因子近似是一个长期的开放问题,但在次指数时间内可以实现恒定的因子近似值(ARXIV:1906.11385)。我们的主要结果意味着PB和$ \ MATHCAL {MSSC} _F $具有相同的属性。我们还研究了一个案例,使价值分布更简洁地作为$ m $产品分布的混合物。这个问题再次与最佳决策树的嘈杂变体有关,该变体更具挑战性。我们给出一个恒定的因子近似值,该近似时间$ n^{\ tilde o(m^2/\ varepsilon^2)} $当每个盒子上的混合组件在电视距离中相同或通过$ \ varepsilon $在电视距离中相同或分开。
translated by 谷歌翻译
最近的经验工作表明,即使所有广告商以非歧视性方式出价,在线广告也可以在用户交付广告时展示偏见。我们研究了广告拍卖的设计,鉴于公平的出价,保证有关展览会产生公平的结果。遵循DWORK和ILVENTO(2019)和CHAWLA等人的作品。 (2020年),我们的目标是设计一种真实的拍卖,这些拍卖会满足其结果的“个人公平”:非正式地说,相似彼此的用户应该获得类似的广告分配。在本框架内,我们量化了社会福利最大化和公平性之间的权衡。这项工作提出了两个概念贡献。首先,我们将公平约束表达为一种稳定条件:所有广告商的任何两个用户都分配了乘法相似的值,必须为每个广告商接受类似的相似分配。该值稳定性约束表示为函数,该函数将值向量之间的乘法距离映射到相应分配之间的最大允许$ \ {\ infty} $距离。标准拍卖不满足这种价值稳定性。其次,我们介绍了一个新的一类分配算法,称为反比例分配,实现公平和社会福利之间的近似最佳权衡,以实现广泛和表现力的价值稳定条件。这些分配算法是真实的,并且先前的,并且实现了最佳(无约会)社会福利的恒定因素近似。特别地,近似比与系统中的广告商的数量无关。在这方面,这些分配算法极大地超越了以前的工作中实现的保证。我们还将结果扩展到更广泛的公平概念,以至于我们称之为公平性。
translated by 谷歌翻译
Deep learning models are being increasingly applied to imbalanced data in high stakes fields such as medicine, autonomous driving, and intelligence analysis. Imbalanced data compounds the black-box nature of deep networks because the relationships between classes may be highly skewed and unclear. This can reduce trust by model users and hamper the progress of developers of imbalanced learning algorithms. Existing methods that investigate imbalanced data complexity are geared toward binary classification, shallow learning models and low dimensional data. In addition, current eXplainable Artificial Intelligence (XAI) techniques mainly focus on converting opaque deep learning models into simpler models (e.g., decision trees) or mapping predictions for specific instances to inputs, instead of examining global data properties and complexities. Therefore, there is a need for a framework that is tailored to modern deep networks, that incorporates large, high dimensional, multi-class datasets, and uncovers data complexities commonly found in imbalanced data (e.g., class overlap, sub-concepts, and outlier instances). We propose a set of techniques that can be used by both deep learning model users to identify, visualize and understand class prototypes, sub-concepts and outlier instances; and by imbalanced learning algorithm developers to detect features and class exemplars that are key to model performance. Our framework also identifies instances that reside on the border of class decision boundaries, which can carry highly discriminative information. Unlike many existing XAI techniques which map model decisions to gray-scale pixel locations, we use saliency through back-propagation to identify and aggregate image color bands across entire classes. Our framework is publicly available at \url{https://github.com/dd1github/XAI_for_Imbalanced_Learning}
translated by 谷歌翻译
We can protect user data privacy via many approaches, such as statistical transformation or generative models. However, each of them has critical drawbacks. On the one hand, creating a transformed data set using conventional techniques is highly time-consuming. On the other hand, in addition to long training phases, recent deep learning-based solutions require significant computational resources. In this paper, we propose PrivateSMOTE, a technique designed for competitive effectiveness in protecting cases at maximum risk of re-identification while requiring much less time and computational resources. It works by synthetic data generation via interpolation to obfuscate high-risk cases while minimizing data utility loss of the original data. Compared to multiple conventional and state-of-the-art privacy-preservation methods on 20 data sets, PrivateSMOTE demonstrates competitive results in re-identification risk. Also, it presents similar or higher predictive performance than the baselines, including generative adversarial networks and variational autoencoders, reducing their energy consumption and time requirements by a minimum factor of 9 and 12, respectively.
translated by 谷歌翻译
Adversarial training has been empirically shown to be more prone to overfitting than standard training. The exact underlying reasons still need to be fully understood. In this paper, we identify one cause of overfitting related to current practices of generating adversarial samples from misclassified samples. To address this, we propose an alternative approach that leverages the misclassified samples to mitigate the overfitting problem. We show that our approach achieves better generalization while having comparable robustness to state-of-the-art adversarial training methods on a wide range of computer vision, natural language processing, and tabular tasks.
translated by 谷歌翻译
Link prediction is a crucial problem in graph-structured data. Due to the recent success of graph neural networks (GNNs), a variety of GNN-based models were proposed to tackle the link prediction task. Specifically, GNNs leverage the message passing paradigm to obtain node representation, which relies on link connectivity. However, in a link prediction task, links in the training set are always present while ones in the testing set are not yet formed, resulting in a discrepancy of the connectivity pattern and bias of the learned representation. It leads to a problem of dataset shift which degrades the model performance. In this paper, we first identify the dataset shift problem in the link prediction task and provide theoretical analyses on how existing link prediction methods are vulnerable to it. We then propose FakeEdge, a model-agnostic technique, to address the problem by mitigating the graph topological gap between training and testing sets. Extensive experiments demonstrate the applicability and superiority of FakeEdge on multiple datasets across various domains.
translated by 谷歌翻译
Adversarial training is widely acknowledged as the most effective defense against adversarial attacks. However, it is also well established that achieving both robustness and generalization in adversarially trained models involves a trade-off. The goal of this work is to provide an in depth comparison of different approaches for adversarial training in language models. Specifically, we study the effect of pre-training data augmentation as well as training time input perturbations vs. embedding space perturbations on the robustness and generalization of BERT-like language models. Our findings suggest that better robustness can be achieved by pre-training data augmentation or by training with input space perturbation. However, training with embedding space perturbation significantly improves generalization. A linguistic correlation analysis of neurons of the learned models reveal that the improved generalization is due to `more specialized' neurons. To the best of our knowledge, this is the first work to carry out a deep qualitative analysis of different methods of generating adversarial examples in adversarial training of language models.
translated by 谷歌翻译
尽管图神经网络(GNNS)已经证明了它们在处理非欧国人结构数据方面的功效,但由于多跳数据依赖性施加的可伸缩性约束,因此很难将它们部署在实际应用中。现有方法试图通过使用训练有素的GNN的标签训练多层感知器(MLP)来解决此可伸缩性问题。即使可以显着改善MLP的性能,但两个问题仍能阻止MLP的表现优于GNN并在实践中使用:图形结构信息的无知和对节点功能噪声的敏感性。在本文中,我们建议在图(NOSMOG)上学习噪声稳定结构感知的MLP,以克服挑战。具体而言,我们首先将节点内容与位置功能进行补充,以帮助MLP捕获图形结构信息。然后,我们设计了一种新颖的表示相似性蒸馏策略,以将结构节点相似性注入MLP。最后,我们介绍了对抗性功能的扩展,以确保稳定的学习能力噪声,并进一步提高性能。广泛的实验表明,在七个数据集中,NOSMOG在转导和归纳环境中均优于GNN和最先进的方法,同时保持竞争性推理效率。
translated by 谷歌翻译
生成的自我监督学习(SSL),尤其是蒙面自动编码器,已成为最令人兴奋的学习范式之一,并且在处理图形数据方面表现出了巨大的潜力。但是,现实世界图总是异质的,它提出了现有方法忽略的三个关键挑战:1)如何捕获复杂的图形结构? 2)如何合并各种节点属性? 3)如何编码不同的节点位置?鉴于此,我们研究了异质图上生成SSL的问题,并提出了HGMAE,这是一种新型的异质图掩盖自动编码器模型,以应对这些挑战。 HGMAE通过两种创新的掩蔽技术和三种独特的培训策略捕获了全面的图形信息。特别是,我们首先使用动态掩模速率开发Metapath掩盖和自适应属性掩蔽,以实现在异质图上有效和稳定的学习。然后,我们设计了几种培训策略,包括基于Metapath的边缘重建,以采用复杂的结构信息,目标属性恢复以结合各种节点属性,以及位置特征预测以编码节点位置信息。广泛的实验表明,HGMAE在多个数据集上的几个任务上均优于对比度和生成的最新基准。
translated by 谷歌翻译
必须密切监控网络物理系统(CPS),以识别并潜在地缓解其常规操作期间出现的新兴问题。但是,他们通常产生的多元时间序列数据可能很复杂,可以理解和分析。虽然正式的产品文档通常会提供诊断建议的示例数据图,但属性,关键阈值和数据交互的纯粹多样性可能会使非专家们不知所措,他们随后从讨论论坛中寻求帮助来解释其数据日志。深度学习模型,例如长期记忆(LSTM)网络,可用于自动化这些任务,并提供对实时多元数据流中检测到的各种异常的明确解释。在本文中,我们介绍了RESAM,该过程是一项需求过程,该过程旨在整合领域专家,讨论论坛和正式产品文档的知识,以以时间序列属性的形式发现和指定需求和设计定义,这些属性有助于构建有效的深度学习异常检测器。我们提出了一个基于针对小型无空天空系统的飞行控制系统的案例研究,并证明其使用指导有效的异常检测模型的构建,同时还为解释性提供了基本支持。 RESAM与开放或关闭的在线论坛为日志分析提供讨论支持的域相关。
translated by 谷歌翻译